CV4Edu - Computer Vision for Education
Computer vision (CV) plays a central role in human‑centered AI, yet most models are trained on web‑scale benchmarks that poorly reflect real classrooms. Educational data are noisy, private, small‑scale, and multimodal (e.g.,face, gaze, pose). Students’ cognitive/behavioral states (e.g.,engagement, mind‑wandering) and learning processes (e.g.,self-regulation, collaboration) can be inferred from subtle cues in the lab. Still, today’s models struggle to generalize to noisy classroom data. CV4Edu brings together computer vision, human-computer interaction, and educational researchers to chart a community agenda for efficient, privacy‑aware multimodal data-driven models that work more efficiently and reliably in low‑resource, real‑world classrooms—potentially launching shared datasets, metrics, and unified practices.
Topics
The workshop topics include (but are not limited to):
Multimodal classroom perception
- Face, gaze, pose, gesture, posture, affect, and prosody
- Video, audio, gaze sensors, and wearables (egocentric and exocentric)
- Multimodal fusion, representation learning, and cross-view / multi-camera setups
Language-centered multimodal learning analytics
- Linking speech/text to video events, gaze/attention, and instructional context
- Classroom NLP: ASR robustness, diarization, evaluating and mitigating bias, discourse modeling, dialogue/tutoring interactions, simplification, misconception detection
- Retrieval-augmented classroom analytics, model adaptation, evaluation for learning-aligned outcomes
Robustness & generalization
- Domain shift beyond the lab, occlusions, noisy data, and missing modalities
- Few-/low-shot learning, continual and on-device adaptation
- Generalization across classroom layouts and populations
Human behavior modeling for learning
- Engagement, attention, affect, confusion, self-regulation, and metacognition
- Collaboration, group dynamics, and teacher–student interactions
- Gaze-informed models, saliency/scanpath prediction, activity recognition
Temporal modeling & intervention
- Sequential/temporal models of learning processes
- Behavioral forecasting, early-warning systems, and interventions
- Real-time inference, feedback, and human-in-the-loop systems
Interpretability, reliability & evaluation
- Interpretable models, uncertainty estimation, and calibration
- OOD detection, fairness, and bias analysis
- Evaluation protocols aligned with learning outcomes
Privacy-aware AI, datasets & deployments
- Privacy-preserving data collection, anonymization, de-identification, and governance
- Annotation strategies, construct-aligned labeling, active learning, synthetic data, and dataset curation
- Classroom-ready systems, scalable multimodal data-collection frameworks, edge/on-device inference, and real-world deployments
We encourage general computer-vision, visually grounded NLP, and human-centered, collaborative AI submissions (e.g., behavioral modeling, pose/activity recognition, gaze estimation, attention modeling, multimodal learning, methods “in the wild”, cognitive state inference and forecasting) that make a clear connection to educational/learning environments (even if primarily in the discussion).
Call for Papers
The workshop invites submissions presenting original research, emerging ideas, datasets and benchmarks, systems, applications, and position papers advancing computer vision for real-world educational settings. We welcome both archival and non-archival contributions, including early-stage work and previously published research, with the goal of fostering discussion and community building.
Submission Tracks
All submissions must follow the CVPR 2026 paper template and official style guidelines.
Archival Track (Full Papers)
Papers submitted to the Archival Track must present original, unpublished work and will be considered for inclusion in the official CVPR 2026 workshop proceedings. The main text must be 6–8 pages in length and formatted according to the CVPR 2026 submission guidelines. References and appendices are not subject to a page limit.Non-Archival Track (Extended Abstracts + Short / Position Papers)
We invite non-archival submissions describing ongoing projects, preliminary results, datasets or benchmarks in progress, negative results, lessons learned, position papers, and work previously published elsewhere (including papers on arXiv or at other venues). These submissions will not be included in the official proceedings. Extended abstracts may be up to 2 pages and short/position papers up to 4 pages (excluding references), formatted according to the CVPR 2026 submission guidelines.Review Process
All submissions will undergo double-blind peer review.
- Archival submissions will receive at least two reviews, followed by a meta-review.
- Submissions must comply with CVPR policies.
- An ethics/IRB checklist is required where applicable, and an optional ethics and broader impact statement may be included.
Important Dates (AoE)
- Paper Submission Deadline (All Tracks): March 12, 2026
- Notification of Decision: April 3, 2026
- Camera-Ready Deadline (Archival Only): April 10, 2026
Submission Site
Papers can be submitted through the OpenReview Submission Site.
Tentative Schedule
Opening & Goals |
|
Keynotes |
|
Poster Session 1 + Coffee Break |
|
Keynotes |
|
Poster Session 2 + Coffee Break |
|
Panel: From Lab to In‑The‑Wild |
|
Community Discussion |
|
Closing & Next Steps |
Venue
700 14th Street
Denver CO 80202
The workshop will be held together with CVPR 2026.
Workshop Organizers
For any questions about the workshop, please contact cv4edu.cvpr@gmail.com
Ekta Sood
Joyce Horn Fonteles
Mariah Bradford
Paul Gavrikov
Prajit Dhar
Janis Pagel
Trisha Mital
Gautam Biswas
Sidney D'Mello